#FP8 training30/10/2025
Ant Group Unveils Ling 2.0 — Scaling Sparse MoE Reasoning to 1T with 1/32 Activation
'Ling 2.0 is a reasoning-first sparse MoE family from Ant Group that keeps per-token compute low with a 1/32 activation recipe while scaling from 16B to 1T parameters.'